Goto

Collaborating Authors

 Poway


Exploring Prime Number Classification: Achieving High Recall Rate and Rapid Convergence with Sparse Encoding

Lee, Serin, Kim, S.

arXiv.org Artificial Intelligence

This paper presents a novel approach at the intersection of machine learning and number theory, focusing on the classification of prime and non-prime numbers. At the core of our research is the development of a highly sparse encoding method, integrated with conventional neural network architectures. This combination has shown promising results, achieving a recall of over 99\% in identifying prime numbers and 79\% for non-prime numbers from an inherently imbalanced sequential series of integers, while exhibiting rapid model convergence before the completion of a single training epoch. We performed training using $10^6$ integers starting from a specified integer and tested on a different range of $2 \times 10^6$ integers extending from $10^6$ to $3 \times 10^6$, offset by the same starting integer. While constrained by the memory capacity of our resources, which limited our analysis to a span of $3\times10^6$, we believe that our study contribute to the application of machine learning in prime number analysis. This work aims to demonstrate the potential of such applications and hopes to inspire further exploration and possibilities in diverse fields.


DeepMPR: Enhancing Opportunistic Routing in Wireless Networks through Multi-Agent Deep Reinforcement Learning

Kaviani, Saeed, Ryu, Bo, Ahmed, Ejaz, Kim, Deokseong, Kim, Jae, Spiker, Carrie, Harnden, Blake

arXiv.org Artificial Intelligence

Opportunistic routing relies on the broadcast capability of wireless networks. It brings higher reliability and robustness in highly dynamic and/or severe environments such as mobile or vehicular ad-hoc networks (MANETs/VANETs). To reduce the cost of broadcast, multicast routing schemes use the connected dominating set (CDS) or multi-point relaying (MPR) set to decrease the network overhead and hence, their selection algorithms are critical. Common MPR selection algorithms are heuristic, rely on coordination between nodes, need high computational power for large networks, and are difficult to tune for network uncertainties. In this paper, we use multi-agent deep reinforcement learning to design a novel MPR multicast routing technique, DeepMPR, which is outperforming the OLSR MPR selection algorithm while it does not require MPR announcement messages from the neighbors. Our evaluation results demonstrate the performance gains of our trained DeepMPR multicast forwarding policy compared to other popular techniques.


DeepADMR: A Deep Learning based Anomaly Detection for MANET Routing

Yahja, Alex, Kaviani, Saeed, Ryu, Bo, Kim, Jae H., Larson, Kevin A.

arXiv.org Artificial Intelligence

We developed DeepADMR, a novel neural anomaly detector for the deep reinforcement learning (DRL)-based DeepCQ+ MANET routing policy. The performance of DRL-based algorithms such as DeepCQ+ is only verified within the trained and tested environments, hence their deployment in the tactical domain induces high risks. DeepADMR monitors unexpected behavior of the DeepCQ+ policy based on the temporal difference errors (TD-errors) in real-time and detects anomaly scenarios with empirical and non-parametric cumulative-sum statistics. The DeepCQ+ design via multi-agent weight-sharing proximal policy optimization (PPO) is slightly modified to enable the real-time estimation of the TD-errors. We report the DeepADMR performance in the presence of channel disruptions, high mobility levels, and network sizes beyond the training environments, which shows its effectiveness.


Deep Reinforcement Learning for System-on-Chip: Myths and Realities

Sung, Tegg Taekyong, Ryu, Bo

arXiv.org Artificial Intelligence

Neural schedulers based on deep reinforcement learning (DRL) have shown considerable potential for solving real-world resource allocation problems, as they have demonstrated significant performance gain in the domain of cluster computing. In this paper, we investigate the feasibility of neural schedulers for the domain of System-on-Chip (SoC) resource allocation through extensive experiments and comparison with non-neural, heuristic schedulers. The key finding is three-fold. First, neural schedulers designed for cluster computing domain do not work well for SoC due to i) heterogeneity of SoC computing resources and ii) variable action set caused by randomness in incoming jobs. Second, our novel neural scheduler technique, Eclectic Interaction Matching (EIM), overcomes the above challenges, thus significantly improving the existing neural schedulers. Specifically, we rationalize the underlying reasons behind the performance gain by the EIM-based neural scheduler. Third, we discover that the ratio of the average processing elements (PE) switching delay and the average PE computation time significantly impacts the performance of neural SoC schedulers even with EIM. Consequently, future neural SoC scheduler design must consider this metric as well as its implementation overhead for practical utility.


The Rise of A.I. Fighter Pilots

The New Yorker

This content can also be viewed on the site it originates from. On a cloudless morning last May, a pilot took off from the Niagara Falls International Airport, heading for restricted military airspace over Lake Ontario. The plane, which bore the insignia of the United States Air Force, was a repurposed Czechoslovak jet, an L-39 Albatros, purchased by a private defense contractor. The bay in front of the cockpit was filled with sensors and computer processors that recorded the aircraft's performance. For two hours, the pilot flew counterclockwise around the lake.


In-flight Novelty Detection with Convolutional Neural Networks

Hartwell, Adam, Montana, Felipe, Jacobs, Will, Kadirkamanathan, Visakan, Mills, Andrew R, Clark, Tom

arXiv.org Artificial Intelligence

Gas turbine engines are complex machines that typically generate a vast amount of data, and require careful monitoring to allow for cost-effective preventative maintenance. In aerospace applications, returning all measured data to ground is prohibitively expensive, often causing useful, high value, data to be discarded. The ability to detect, prioritise, and return useful data in real-time is therefore vital. This paper proposes that system output measurements, described by a convolutional neural network model of normality, are prioritised in real-time for the attention of preventative maintenance decision makers. Due to the complexity of gas turbine engine time-varying behaviours, deriving accurate physical models is difficult, and often leads to models with low prediction accuracy and incompatibility with real-time execution. Data-driven modelling is a desirable alternative producing high accuracy, asset specific models without the need for derivation from first principles. We present a data-driven system for online detection and prioritisation of anomalous data. Biased data assessment deriving from novel operating conditions is avoided by uncertainty management integrated into the deep neural predictive model. Testing is performed on real and synthetic data, showing sensitivity to both real and synthetic faults. The system is capable of running in real-time on low-power embedded hardware and is currently in deployment on the Rolls-Royce Pearl 15 engine flight trials.


Local Model Feature Transformations

Brown, CScott

arXiv.org Machine Learning

Local learning methods are a popular class of machine learning algorithms. The basic idea for the entire cadre is to choose some non-local model family, to train many of them on small sections of neighboring data, and then to `stitch' the resulting models together in some way. Due to the limits of constraining a training dataset to a small neighborhood, research on locally-learned models has largely been restricted to simple model families. Also, since simple model families have no complex structure by design, this has limited use of the individual local models to predictive tasks. We hypothesize that, using a sufficiently complex local model family, various properties of the individual local models, such as their learned parameters, can be used as features for further learning. This dissertation improves upon the current state of research and works toward establishing this hypothesis by investigating algorithms for localization of more complex model families and by studying their applications beyond predictions as a feature extraction mechanism. We summarize this generic technique of using local models as a feature extraction step with the term ``local model feature transformations.'' In this document, we extend the local modeling paradigm to Gaussian processes, orthogonal quadric models and word embedding models, and extend the existing theory for localized linear classifiers. We then demonstrate applications of local model feature transformations to epileptic event classification from EEG readings, activity monitoring via chest accelerometry, 3D surface reconstruction, 3D point cloud segmentation, handwritten digit classification and event detection from Twitter feeds.


Biomimicry Gives a Lift to AI in Aviation – Northrop Grumman

#artificialintelligence

Who among us hasn't stared up at a hawk or a vulture circling lazily in the sky and wondered how they stay aloft so long? Or wondered how sky-darkening flocks of migrating birds can travel thousands of miles so quickly and so effortlessly? Researchers are now able to tackle these questions more systematically using biomimicry, the process of imitating nature's systems to solve complex human problems. Their goal is to develop an artificial intelligence algorithm that will allow gliders -- either piloted or autonomous sailplanes -- to mimic the flight behavior and endurance of birds. Soon, AI in aviation will transform the types of missions that gliders perform.


AI gliders learn to fly using air currents, just like birds

#artificialintelligence

Birds don't always flap their wings to fly; sometimes they soar by taking advantage of rising columns of warm air known as thermals. With large wingspans, they can stay aloft for hours while expending minimal energy. Exactly how they do it -- navigating tiny changes in unpredictable air currents -- isn't well-known. But scientists are now using artificial intelligence to learn their tricks, and hopefully, they can teach our aircraft to do the same. As described in a paper published this week in the journal Nature, researchers from universities in the US and Italy used machine learning to train an algorithm to control a glider to navigate thermals.


AI gliders learn to fly using air currents, just like birds

#artificialintelligence

Birds don't always flap their wings to fly; sometimes they soar by taking advantage of rising columns of warm air known as thermals. With large wingspans, they can stay aloft for hours while expending minimal energy. Exactly how they do it -- navigating tiny changes in unpredictable air currents -- isn't well-known. But scientists are now using artificial intelligence to learn their tricks, and hopefully, they can teach our aircraft to do the same. As described in a paper published this week in the journal Nature, researchers from universities in the US and Italy used machine learning to train an algorithm to control a glider to navigate thermals. It's not the first time artificial intelligence has been used for this task (Microsoft published similar work with gliders last year), but it's the first time that data from real flights has been used to update and improve an AI's performance in the field.